Toward a Multi-Speaker Visual Articulatory Feedback System

نویسندگان

  • Atef Ben Youssef
  • Thomas Hueber
  • Pierre Badin
  • Gérard Bailly
چکیده

In this paper, we present recent developments on the HMMbased acoustic-to-articulatory inversion approch that we develop for a “visual articulatory feedback” system. In this approach, multi-stream phoneme HMMs are trained jointly on synchronous streams of acoustic and articulatory data, acquired by electromagnetic articulography (EMA). Acousticto-articulatory inversion is achieved in two steps. Phonetic and state decoding is first performed. Then articulatory trajectories are inferred from the decoded phone and state sequence using the maximum-likelihood parameter generation algorithm (MLPG). We introduce here a new procedure for the reestimation of the HMM parameters, based on the Minimum Generation Error criterion (MGE). We also investigate the use of model adaptation techniques based on maximum likelihood linear regression (MLLR), as a first step toward a multispeaker visual articulatory feedback system.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Toward a speaker-independent visual articulatory feedback system

Context Several studies tend to show that visual articulatory feedback is useful for phonetic correction, both for speech therapy and “Computer Aided Pronunciation Training” (CAPT) [1]. In [2], we proposed a visual articulatory feedback system based on a 3D talking head used in “an augmented speech scenario”, i.e. displaying all speech articulators including the tongue and velum. In the propose...

متن کامل

Automatic animation of an articulatory tongue model from ultrasound images using Gaussian mixture regression

This paper presents a method for automatically animating the articulatory tongue model of a reference speaker from ultrasound images of the tongue of another speaker. This work is developed in the context of speech therapy based on visual biofeedback, where a speaker is provided with visual information about his/her own articulation. In our approach, the feedback is delivered via an articulator...

متن کامل

Cross-speaker Acoustic-to-Articulatory Inversion using Phone-based Trajectory HMM for Pronunciation Training

The article presents a statistical mapping approach for crossspeaker acoustic-to-articulatory inversion. The goal is to estimate the most likely articulatory trajectories for a reference speaker from the speech audio signal of another speaker. This approach is developed in the framework of our system of visual articulatory feedback developed for computer-assisted pronunciation training applicat...

متن کامل

Pronunciation analysis by acoustic-to-articulatory feature inversion

Second language learners may require assistance correcting their articulation of unfamiliar phonemes in order to reach the target pronunciation. If, e.g., a talking head is to provide the learner with feedback on how to change the articulation, a required first step is to be able to analyze the learner’s articulation. This paper describes how a specialized restricted acoustic-to-articulatory in...

متن کامل

Speaker adaptation of an acoustic-articulatory inversion model using cascaded Gaussian mixture regressions

The article presents a method for adapting a GMM-based acoustic-articulatory inversion model trained on a reference speaker to another speaker. The goal is to estimate the articulatory trajectories in the geometrical space of a reference speaker from the speech audio signal of another speaker. This method is developed in the context of a system of visual biofeedback, aimed at pronunciation trai...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2011